Learning Kernel-Based Halfspaces with the 0-1 Loss
نویسندگان
چکیده
We describe and analyze a new algorithm for agnostically learning kernel-based halfspaces with respect to the 0-1 loss function. Unlike most of the previous formulations, which rely on surrogate convex loss functions (e.g., hinge-loss in support vector machines (SVMs) and log-loss in logistic regression), we provide finite time/sample guarantees with respect to the more natural 0-1 loss function. The proposed algorithm can learn kernel-based halfspaces in worst-case time poly(exp(L log(L/ ))), for any distribution, where L is a Lipschitz constant (which can be thought of as the reciprocal of the margin), and the learned classifier is worse than the optimal halfspace by at most . We also prove a hardness result, showing that under a certain cryptographic assumption, no algorithm can learn kernel-based halfspaces in time polynomial in L.
منابع مشابه
Learning Kernel-Based Halfspaces with the Zero-One Loss
We describe and analyze a new algorithm for agnostically learning kernel-based halfspaces with respect to the zero-one loss function. Unlike most previous formulations which rely on surrogate convex loss functions (e.g. hinge-loss in SVM and log-loss in logistic regression), we provide finite time/sample guarantees with respect to the more natural zero-one loss function. The proposed algorithm ...
متن کاملLearning Halfspaces and Neural Networks with Random Initialization
We study non-convex empirical risk minimization for learning halfspaces and neural networks. For loss functions that are L-Lipschitz continuous, we present algorithms to learn halfspaces and multi-layer neural networks that achieve arbitrarily small excess risk > 0. The time complexity is polynomial in the input dimension d and the sample size n, but exponential in the quantity (L/ ) log(L/ ). ...
متن کاملAgnostically Learning Halfspaces with Margin Errors
We describe and analyze a new algorithm for agnostically learning half-spaces with respect to the margin error rate. Roughly speaking, this corre-sponds to the worst-case error rate after each point is perturbed by a noisevector of length at most μ. Margin based analysis is widely used in learningtheory and is considered the most successful theoretical explanation for thesta...
متن کاملLearning Intersections of Halfspaces with a Margin
We give a new algorithm for learning intersections of halfspaces with a margin, i.e. under the assumption that no example lies too close to any separating hyperplane. Our algorithm combines random projection techniques for dimensionality reduction, polynomial threshold function constructions, and kernel methods. The algorithm is fast and simple. It learns a broader class of functions and achiev...
متن کاملLearning Functions of Halfspaces using Prefix Covers
We present a simple query-algorithm for learning arbitrary functions of k halfspaces under any product distribution on the Boolean hypercube. Our algorithms learn any function of k halfspaces to within accuracy ε in time O((nk/ε)) under any product distribution on {0, 1} using read-once branching programs as a hypothesis.. This gives the first poly(n, 1/ε) algorithm for learning even the inters...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- SIAM J. Comput.
دوره 40 شماره
صفحات -
تاریخ انتشار 2011